# Efficient Pretraining
Genmedclip
MIT
GenMedClip is a zero-shot image classification model based on the open_clip library, specializing in medical image analysis.
Image Classification
G
wisdomik
40
0
Rho Math 1b V0.1
MIT
Rho-1 is a language model specialized in mathematics, pretrained with Selective Language Modeling (SLM) method, significantly improving accuracy in solving mathematical problems.
Large Language Model
Transformers English

R
microsoft
1,451
15
Tinyllama V1.1 Math Code
Apache-2.0
TinyLlama is a compact language model with 1.1 billion parameters, adopting the same architecture and tokenizer as Llama 2, suitable for applications with limited computational and memory resources.
Large Language Model
Transformers English

T
TinyLlama
3,436
11
Tinyllama 1.1B Intermediate Step 1431k 3T
Apache-2.0
TinyLlama is a 1.1B parameter Llama model pretrained on 3 trillion tokens, designed to provide compact and efficient text generation capabilities.
Large Language Model
Transformers English

T
TinyLlama
25.04k
173
Tinyllama 1.1B Intermediate Step 1195k Token 2.5T
Apache-2.0
TinyLlama is a compact 1.1B-parameter Llama model pretrained on 3 trillion tokens, designed for resource-constrained environments.
Large Language Model
Transformers English

T
TinyLlama
419
52
Sheared LLaMA 2.7B
Apache-2.0
Sheared-LLaMA-2.7B is a lightweight language model derived from Llama-2-7b through pruning and continued pretraining, consuming only a 50B token budget.
Large Language Model
Transformers

S
princeton-nlp
1,131
60
Tinyllama 1.1B Step 50K 105b
Apache-2.0
TinyLlama is a 1.1B parameter Llama model, planned to be pretrained on 3 trillion tokens, optimized to complete training in 90 days on 16 A100-40G GPUs.
Large Language Model
Transformers English

T
TinyLlama
14.41k
133
Codet5p 16b
Bsd-3-clause
CodeT5+ 16B is an open-source family of large language models for code, featuring an encoder-decoder architecture that supports multiple modes, suitable for a wide range of code understanding and generation tasks.
Large Language Model
Transformers

C
Salesforce
292
65
Videomae Small Finetuned Kinetics
VideoMAE is a masked autoencoder model for video, pretrained with self-supervision and fine-tuned on the Kinetics-400 dataset, suitable for video classification tasks.
Video Processing
Transformers

V
MCG-NJU
2,152
1
Videomae Huge Finetuned Kinetics
VideoMAE is a video pretraining model based on Masked Autoencoder (MAE), fine-tuned on the Kinetics-400 dataset through self-supervised learning, suitable for video classification tasks.
Video Processing
Transformers

V
MCG-NJU
2,984
4
Indobertweet Base Uncased
Apache-2.0
The first pre-trained language model specifically for Indonesian Twitter, built by extending Indonesian BERT with domain-specific vocabulary
Large Language Model
Transformers Other

I
indolem
2,848
12
Arabictransformer Base
An efficient Arabic language model based on Funnel Transformer and ELECTRA objective, with low computational cost and superior performance
Large Language Model
Transformers

A
sultan
17
1
Bertin Roberta Base Spanish
BERTIN is a series of Spanish BERT-based models. The current model is a RoBERTa-base model trained from scratch on a portion of the Spanish mC4 dataset using Flax.
Large Language Model Spanish
B
bertin-project
1,845
36
Bert Base Uncased Sparse 90 Unstructured Pruneofa
Apache-2.0
This is a sparsely pretrained BERT-Base model achieving 90% weight sparsity through one-shot pruning, suitable for fine-tuning on various language tasks.
Large Language Model
Transformers English

B
Intel
178
0
Featured Recommended AI Models